-
Notifications
You must be signed in to change notification settings - Fork 313
feat: Add an option for specifying model name #685
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
@alvarobartt Let me know if you have are any changes/feedback for this PR |
@alvarobartt @Narsil I was wondering if you have any thoughts/concerns/feedback regarding the use-case/implementation for this PR! Let me know if this is something you would like to discuss offline and I am open to that as well! |
Hey @vrdn-23 thanks for opening this PR and apologies I'm just looking into it now! But I'll check that everything works this week, and happy to support it and add it within the next release 🤗 |
Thanks @alvarobartt for getting back to me. let me know if there are any changes I need to make! |
@alvarobartt just wanted to check in and see if this was still on your radar for review! |
Hey @vrdn-23, yes! This is something I'd like to include for Text Embeddings Inference v1.9.0, but I'd like to first make sure that some patches land, apologies for the delay 🙏🏻 Also given that we add the |
I think that would be great! I think by default if the model name isn't specified in vLLM, it still serves the request, so we shouldn't have to change compatibility of the OpenAI format spec. I can add a check for the validation based on the model name specified in |
Just to add to this, it does seem vLLM does not check for matching with
|
Yes that's right @vrdn-23, I'll validate and approve ✅ |
What does this PR do?
This PR introduces a new CLI argument,
--served-model-name
, which allows users to specify a custom model name to be returned in responses from the OpenAI-compatible endpoint.This is particularly useful in scenarios where the model is loaded from a local path (e.g., /data/model) and does not have an inherent name associated with it. By setting
--served-model-name
, users can override the default model identifier (which might be a generic or filesystem-based value) and provide a more descriptive or meaningful name in the API response. This helps improve clarity and consistency, especially when integrating with clients or tools that rely on the model field in the response for tracking or routing purposes.Before submitting
Pull Request section?
to it if that's the case.
documentation guidelines, and
here are tips on formatting docstrings.
Who can review?
@Narsil @alvarobartt @kozistr
I have tested this to the best of my ability, but I'm not sure if I did the gRPC bits correctly, so if someone could help verify that, that would be great!